Classifier learning with a new locality regularization method

نویسندگان

  • Hui Xue
  • Songcan Chen
  • Xiaoqin Zeng
چکیده

It is well known that the generalization capability is one of the most important criterions to develop and evaluate a classifier for a given pattern classification problem. The localized generalization error model (RSM) [2, 12] recently proposed by Ng et al. provides a more intuitive look at the generalization error. Although RSM gives a brand-new method to promote the generalization performance, it is in nature equivalent to another type of regularization. In this paper, we first prove the essential relationship between RSM and regularization, and demonstrate that the stochastic sensitivity measure in RSM exactly corresponds to a regularizing term. Then, we develop a new generalization error bound from the regulation viewpoint, which is inspired by the proved relationship between RSM and regularization. Moreover, we derive a new regularization method, called as locality regularization (LR), from the bound. Different from the existing regularization methods which artificially and externally append the regularizing term in order to smooth the solution, LR is naturally and internally deduced from the defined expected risk functional and calculated by employing locality information. Through combining with spectral graph theory, LR introduces the local structure information of the samples into the regularizing term and further improves the generalization capability. In contrast with RSM, which is relatively * Corresponding author: Tel: +86-25-84896481 Ext. 12106; Fax:+86-25-84498069; E-mail: [email protected] (S. Chen) [email protected] (H. Xue), and [email protected](X. Zeng) sensitive to the different sampling of the samples, LR uses the discrete k-neighborhood rather than the common continuous Q-neighborhood in RSM to differentiate the relative position of different training samples automatically and avoid the complex computation of Q for various classifiers. Furthermore, LR uses the regularization parameter to control the trade-off between the training accuracy and the classifier stability. Experimental results on artificial and real world problems show that LR yields better generalization capability than both RSM and some traditional regularization methods.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

L1/Lp Regularization of Differences

In this paper, we introduce L1/Lp regularization of differences as a new regularization approach that can directly regularize models such as the naive Bayes classifier and (autoregressive) hidden Markov models. An algorithm is developed that selects values of the regularization parameter based on a derived stability condition. for the regularized naive Bayes classifier, we show that the method ...

متن کامل

NEW CRITERIA FOR RULE SELECTION IN FUZZY LEARNING CLASSIFIER SYSTEMS

Designing an effective criterion for selecting the best rule is a major problem in theprocess of implementing Fuzzy Learning Classifier (FLC) systems. Conventionally confidenceand support or combined measures of these are used as criteria for fuzzy rule evaluation. In thispaper new entities namely precision and recall from the field of Information Retrieval (IR)systems is adapted as alternative...

متن کامل

Locality Preserving Semi-Supervised Support Vector Machine

Manifold regularization, which learns from a limited number of labeled samples and a large number of unlabeled samples, is a powerful semi-supervised classifier with a solid theoretical foundation. However, manifold regularization has the tendency to misclassify data near the boundaries of different classes during the classification process. In this paper, we propose a novel classification meth...

متن کامل

Linearized Bregman for l1-regularized Logistic Regression

Sparse logistic regression is an important linear classifier in statistical learning, providing an attractive route for feature selection. A popular approach is based on minimizing an l1-regularization term with a regularization parameter λ that affects the solution sparsity. To determine an appropriate value for the regularization parameter, one can apply the grid search method or the Bayesian...

متن کامل

A Mathematical Analysis of New L-curve to Estimate the Parameters of Regularization in TSVD Method

A new technique to find the optimization parameter in TSVD regularization method is based on a curve which is drawn against the residual norm [5]. Since the TSVD regularization is a method with discrete regularization parameter, then the above-mentioned curve is also discrete. In this paper we present a mathematical analysis of this curve, showing that the curve has L-shaped path very similar t...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • Pattern Recognition

دوره 41  شماره 

صفحات  -

تاریخ انتشار 2008